celeb-a dataset
DCGAN from Scratch with Tensorflow Keras -- Create Fake Images from CELEB-A Dataset
Generator: the generator generates new data instances that are "similar" to the training data, in our case celebA images. Generator takes random latent vector and outputs a "fake" image of the same size as our reshaped celebA image. Discriminator: the discriminator evaluate the authenticity of provided images; it classifies the images from the generator and the original image. Discriminator takes true of fake images and outputs the probability estimate ranging between 0 and 1. Here, D refers to the discriminator network, while G obviously refers to the generator.
- Law > Criminal Law (0.83)
- Information Technology > Security & Privacy (0.83)
Compressing GANs using Knowledge Distillation
Aguinaldo, Angeline, Chiang, Ping-Yeh, Gain, Alex, Patil, Ameya, Pearson, Kolten, Feizi, Soheil
Generative Adversarial Networks (GANs) have been used in several machine learning tasks such as domain transfer, super resolution, and synthetic data generation. State-of-the-art GANs often use tens of millions of parameters, making them expensive to deploy for applications in low SWAP (size, weight, and power) hardware, such as mobile devices, and for applications with real time capabilities. There has been no work found to reduce the number of parameters used in GANs. Therefore, we propose a method to compress GANs using knowledge distillation techniques, in which a smaller "student" GAN learns to mimic a larger "teacher" GAN. We show that the distillation methods used on MNIST, CIFAR-10, and Celeb-A datasets can compress teacher GANs at ratios of 1669:1, 58:1, and 87:1, respectively, while retaining the quality of the generated image. From our experiments, we observe a qualitative limit for GAN's compression. Moreover, we observe that, with a fixed parameter budget, compressed GANs outperform GANs trained using standard training methods. We conjecture that this is partially owing to the optimization landscape of over-parameterized GANs which allows efficient training using alternating gradient descent. Thus, training an over-parameterized GAN followed by our proposed compression scheme provides a high quality generative model with a small number of parameters.
- North America > United States > New York > New York County > New York City (0.04)
- North America > United States > Maryland (0.04)
- Europe > United Kingdom > England > Cambridgeshire > Cambridge (0.04)